This is a Mixture of Experts (MoE) model based on Qwen3, with a total of 2.4 billion parameters, including four expert models with 0.6 billion parameters each, designed to deliver more accurate results with higher efficiency and lower memory usage.
Large Language Model
Transformers Supports Multiple Languages